首页> 外文OA文献 >DropIn: Making reservoir computing neural networks robust to missing inputs by dropout
【2h】

DropIn: Making reservoir computing neural networks robust to missing inputs by dropout

机译:DropIn:使储层计算神经网络对因丢失而丢失的输入具有鲁棒性

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

The paper presents a novel, principled approach to train recurrent neural networks from the Reservoir Computing family that are robust to missing part of the input features at prediction time. By building on the ensembling properties of Dropout regularization, we propose a methodology, named DropIn, which efficiently trains a neural model as a committee machine of subnetworks, each capable of predicting with a subset of the original input features. We discuss the application of the DropIn methodology in the context of Reservoir Computing models and targeting applications characterized by input sources that are unreliable or prone to be disconnected, such as in pervasive wireless sensor networks and ambient intelligence. We provide an experimental assessment using real-world data from such application domains, showing how the Dropin methodology allows to maintain predictive performances comparable to those of a model without missing features, even when 20%-50% of the inputs are not available.
机译:本文提出了一种新颖的,有原则的方法来训练来自Reservoir Computing系列的递归神经网络,该方法对于预测时缺少部分输入特征具有鲁棒性。通过基于Dropout正则化的集合特性,我们提出了一种名为DropIn的方法,该方法可以有效地将神经模型训练为子网的委员会机器,每个模型都可以使用原始输入特征的子集进行预测。我们讨论了DropIn方法在储层计算模型中的应用以及以输入源为特征的应用,这些输入源的特征是不可靠或易于断开,例如在普及的无线传感器网络和环境智能中。我们使用来自这些应用程序域的真实数据提供了一项实验评估,显示了即使没有20%-50%的输入时,Dropin方法如何能够保持与模型可比的预测性能,而不会丢失功能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号